Coral Sea
Newly discovered deep-sea lanternshark glows in the waters near Australia
The tiny shark and a ghost-like crab are two of the latest species uncovered in a yearslong expedition. Breakthroughs, discoveries, and DIY tips sent every weekday. Oceanographers scouring the waters off of Western Australia have discovered two new deep-sea oddities . On October 6, Australia's Commonwealth Scientific and Industrial Research Organization (CSIRO) showcased these new species originally collected in 2022: a bioluminescent lanternshark and a tiny, semi-translucent porcelain crab . The team revealed two of its initial finds--the painted hornshark and the ridged-egg catshark --in 2023.
- Oceania > Australia > Western Australia (0.25)
- South America > Chile (0.05)
- Pacific Ocean > South Pacific Ocean > Coral Sea (0.05)
- (10 more...)
Mechanistic Interpretability with SAEs: Probing Religion, Violence, and Geography in Large Language Models
Simbeck, Katharina, Mahran, Mariam
Despite growing research on bias in large language models (LLMs), most work has focused on gender and race, with little attention to religious identity. This paper explores how religion is internally represented in LLMs and how it intersects with concepts of violence and geography. Using mechanistic interpretability and Sparse Autoencoders (SAEs) via the Neuronpedia API, we analyze latent feature activations across five models. We measure overlap between religion- and violence-related prompts and probe semantic patterns in activation contexts. While all five religions show comparable internal cohesion, Islam is more frequently linked to features associated with violent language. In contrast, geographic associations largely reflect real-world religious demographics, revealing how models embed both factual distributions and cultural stereotypes. These findings highlight the value of structural analysis in auditing not just outputs but also internal representations that shape model behavior.
- North America > United States > New York > New York County > New York City (0.28)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- Asia > Middle East > Palestine > Gaza Strip > Gaza Governorate > Gaza (0.14)
- (225 more...)
SCAWaveNet: A Spatial-Channel Attention-Based Network for Global Significant Wave Height Retrieval
Zhang, Chong, Liu, Xichao, Zhan, Yibing, Tao, Dapeng, Ni, Jun, Bu, Jinwei
Recent advancements in spaceborne GNSS missions have produced extensive global datasets, providing a robust basis for deep learning-based significant wave height (SWH) retrieval. While existing deep learning models predominantly utilize CYGNSS data with four-channel information, they often adopt single-channel inputs or simple channel concatenation without leveraging the benefits of cross-channel information interaction during training. To address this limitation, a novel spatial-channel attention-based network, namely SCAWaveNet, is proposed for SWH retrieval. Specifically, features from each channel of the DDMs are modeled as independent attention heads, enabling the fusion of spatial and channel-wise information. For auxiliary parameters, a lightweight attention mechanism is designed to assign weights along the spatial and channel dimensions. The final feature integrates both spatial and channel-level characteristics. Model performance is evaluated using four-channel CYGNSS data. When ERA5 is used as a reference, SCAWaveNet achieves an average RMSE of 0.438 m. When using buoy data from NDBC, the average RMSE reaches 0.432 m. Compared to state-of-the-art models, SCAWaveNet reduces the average RMSE by at least 3.52% on the ERA5 dataset and by 5.68% on the NDBC buoy observations. The code is available at https://github.com/Clifx9908/SCAWaveNet.
- Europe > Austria > Vienna (0.14)
- Asia > China > Yunnan Province > Kunming (0.04)
- Pacific Ocean > South Pacific Ocean > Coral Sea (0.04)
- (8 more...)
Do BERT-Like Bidirectional Models Still Perform Better on Text Classification in the Era of LLMs?
Zhang, Junyan, Huang, Yiming, Liu, Shuliang, Gao, Yubo, Hu, Xuming
The rapid adoption of LLMs has overshadowed the potential advantages of traditional BERT-like models in text classification. This study challenges the prevailing "LLM-centric" trend by systematically comparing three category methods, i.e., BERT-like models fine-tuning, LLM internal state utilization, and zero-shot inference across six high-difficulty datasets. Our findings reveal that BERT-like models often outperform LLMs. We further categorize datasets into three types, perform PCA and probing experiments, and identify task-specific model strengths: BERT-like models excel in pattern-driven tasks, while LLMs dominate those requiring deep semantics or world knowledge. Based on this, we propose TaMAS, a fine-grained task selection strategy, advocating for a nuanced, task-driven approach over a one-size-fits-all reliance on LLMs.
RankCoT: Refining Knowledge for Retrieval-Augmented Generation through Ranking Chain-of-Thoughts
Wu, Mingyan, Liu, Zhenghao, Yan, Yukun, Li, Xinze, Yu, Shi, Zeng, Zheni, Gu, Yu, Yu, Ge
Retrieval-Augmented Generation (RAG) enhances the performance of Large Language Models (LLMs) by incorporating external knowledge. However, LLMs still encounter challenges in effectively utilizing the knowledge from retrieved documents, often being misled by irrelevant or noisy information. To address this issue, we introduce RankCoT, a knowledge refinement method that incorporates reranking signals in generating CoT-based summarization for knowledge refinement based on given query and all retrieval documents. During training, RankCoT prompts the LLM to generate Chain-of-Thought (CoT) candidates based on the query and individual documents. It then fine-tunes the LLM to directly reproduce the best CoT from these candidate outputs based on all retrieved documents, which requires LLM to filter out irrelevant documents during generating CoT-style summarization. Additionally, RankCoT incorporates a self-reflection mechanism that further refines the CoT outputs, resulting in higher-quality training data. Our experiments demonstrate the effectiveness of RankCoT, showing its superior performance over other knowledge refinement models. Further analysis reveals that RankCoT can provide shorter but effective refinement results, enabling the generator to produce more accurate answers. All code and data are available at https://github.com/NEUIR/RankCoT.
- Oceania > Australia > Western Australia (0.14)
- Indian Ocean (0.05)
- Pacific Ocean > South Pacific Ocean > Coral Sea (0.04)
- (13 more...)
Building a Taiwanese Mandarin Spoken Language Model: A First Attempt
Yang, Chih-Kai, Fu, Yu-Kuan, Li, Chen-An, Lin, Yi-Cheng, Lin, Yu-Xiang, Chen, Wei-Chih, Chung, Ho Lam, Kuan, Chun-Yi, Huang, Wei-Ping, Lu, Ke-Han, Lin, Tzu-Quan, Wang, Hsiu-Hsuan, Hu, En-Pei, Hsu, Chan-Jan, Tseng, Liang-Hsuan, Chiu, I-Hsiang, Sanga, Ulin, Chen, Xuanjun, Hsu, Po-chun, Yang, Shu-wen, Lee, Hung-yi
This technical report presents our initial attempt to build a spoken large language model (LLM) for Taiwanese Mandarin, specifically tailored to enable real-time, speech-to-speech interaction in multi-turn conversations. Our end-to-end model incorporates a decoder-only transformer architecture and aims to achieve seamless interaction while preserving the conversational flow, including full-duplex capabilities allowing simultaneous speaking and listening. The paper also details the training process, including data preparation with synthesized dialogues and adjustments for real-time interaction. We also developed a platform to evaluate conversational fluency and response coherence in multi-turn dialogues. We hope the release of the report can contribute to the future development of spoken LLMs in Taiwanese Mandarin.
- Pacific Ocean > South Pacific Ocean > Coral Sea (0.04)
- Asia > Thailand > Bangkok > Bangkok (0.04)
- Leisure & Entertainment (1.00)
- Media > Film (0.67)
- Education (0.67)
- Health & Medicine > Consumer Health (0.46)
Improved Forecasts of Global Extreme Marine Heatwaves Through a Physics-guided Data-driven Approach
Shu, Ruiqi, Wu, Hao, Gao, Yuan, Xu, Fanghua, Gou, Ruijian, Huang, Xiaomeng
The unusually warm sea surface temperature events known as marine heatwaves (MHWs) have a profound impact on marine ecosystems. Accurate prediction of extreme MHWs has significant scientific and financial worth. However, existing methods still have certain limitations, especially in the most extreme MHWs. In this study, to address these issues, based on the physical nature of MHWs, we created a novel deep learning neural network that is capable of accurate 10-day MHW forecasting. Our framework significantly improves the forecast ability of extreme MHWs through two specially designed modules inspired by numerical models: a coupler and a probabilistic data argumentation. The coupler simulates the driving effect of atmosphere on MHWs while the probabilistic data argumentation approaches significantly boost the forecast ability of extreme MHWs based on the idea of ensemble forecast. Compared with traditional numerical prediction, our framework has significantly higher accuracy and requires fewer computational resources. What's more, explainable AI methods show that wind forcing is the primary driver of MHW evolution and reveal its relation with air-sea heat exchange. Overall, our model provides a framework for understanding MHWs' driving processes and operational forecasts in the future.
- Pacific Ocean > North Pacific Ocean > South China Sea (0.04)
- Oceania > Australia > Western Australia (0.04)
- Pacific Ocean > South Pacific Ocean > Coral Sea (0.04)
- (6 more...)
Inductive-Deductive Strategy Reuse for Multi-Turn Instructional Dialogues
Ou, Jiao, Wu, Jiayu, Liu, Che, Zhang, Fuzheng, Zhang, Di, Gai, Kun
Aligning large language models (LLMs) with human expectations requires high-quality instructional dialogues, which can be achieved by raising diverse, in-depth, and insightful instructions that deepen interactions. Existing methods target instructions from real instruction dialogues as a learning goal and fine-tune a user simulator for posing instructions. However, the user simulator struggles to implicitly model complex dialogue flows and pose high-quality instructions. In this paper, we take inspiration from the cognitive abilities inherent in human learning and propose the explicit modeling of complex dialogue flows through instructional strategy reuse. Specifically, we first induce high-level strategies from various real instruction dialogues. These strategies are applied to new dialogue scenarios deductively, where the instructional strategies facilitate high-quality instructions. Experimental results show that our method can generate diverse, in-depth, and insightful instructions for a given dialogue history. The constructed multi-turn instructional dialogues can outperform competitive baselines on the downstream chat model.
- Oceania > Australia > Victoria > Melbourne (0.14)
- Oceania > Australia > New South Wales > Sydney (0.14)
- North America > United States > Minnesota > Hennepin County > Minneapolis (0.14)
- (18 more...)
- Research Report > New Finding (0.48)
- Research Report > Experimental Study (0.46)
- Transportation > Infrastructure & Services > Airport (1.00)
- Transportation > Air (1.00)
- Media (1.00)
- (5 more...)
Can the power of artificial intelligence be harnessed help to predict Australia's weather?
Kerry Plowright had his feet up and was watching TV one evening late last year when his phone warned of incoming hail. "I was stunned when I walked out the door because there was just this roar," he says, describing the sound of hailstones hitting roofs in the New South Wales town of Kingscliff. He had just enough time to move his cars under canvas sails, sparing them from damage. This season may include a second tropical cyclone to strike Queensland. The Albanese government has launched an inquiry into warnings issued by the Bureau of Meteorology and emergency authorities after complaints by councils and others that some alerts lacked accuracy and timeliness.
- Oceania > Australia > Queensland (0.26)
- Oceania > Australia > New South Wales (0.25)
- Pacific Ocean > South Pacific Ocean > Coral Sea (0.05)
- (3 more...)
Dataset Distillation with Convexified Implicit Gradients
Loo, Noel, Hasani, Ramin, Lechner, Mathias, Rus, Daniela
We propose a new dataset distillation algorithm using reparameterization and convexification of implicit gradients (RCIG), that substantially improves the state-of-the-art. To this end, we first formulate dataset distillation as a bi-level optimization problem. Then, we show how implicit gradients can be effectively used to compute meta-gradient updates. We further equip the algorithm with a convexified approximation that corresponds to learning on top of a frozen finite-width neural tangent kernel. Finally, we improve bias in implicit gradients by parameterizing the neural network to enable analytical computation of final-layer parameters given the body parameters. RCIG establishes the new state-of-the-art on a diverse series of dataset distillation tasks. Notably, with one image per class, on resized ImageNet, RCIG sees on average a 108\% improvement over the previous state-of-the-art distillation algorithm. Similarly, we observed a 66\% gain over SOTA on Tiny-ImageNet and 37\% on CIFAR-100.
- Pacific Ocean > South Pacific Ocean > Coral Sea (0.04)
- North America > United States > Wisconsin > Dane County > Madison (0.04)
- North America > United States > New York > New York County > New York City (0.04)
- (8 more...)
- Transportation > Ground > Road (1.00)
- Transportation > Passenger (0.69)
- Information Technology > Artificial Intelligence > Representation & Reasoning (1.00)
- Information Technology > Artificial Intelligence > Machine Learning > Statistical Learning > Gradient Descent (0.48)
- Information Technology > Artificial Intelligence > Machine Learning > Neural Networks > Deep Learning (0.46)